A Parallel PARAFAC Implementation & Scalability Testing for Large-Scale Dense Tensor Decomposition
نویسنده
چکیده
Parallel Factor Analysis (PARAFAC) is used in many scientific disciplines to decompose multidimensional datasets into principal factors in order to uncover relationships in the data. While quite popular, the common implementations of PARAFAC are single server solutions that do not scale well to very large datasets. To address this limitation, a Parallel PARAFAC algorithm has been designed and implemented in C using MPI. The end-to-end pipeline includes a parallel read of the input data from a file, the execution of the parallel algorithm, and concludes with a parallel write of the results to a file. The implementation has been evaluated using a strong scaling study on an IBM Blue Gene/Q supercomputer. The compute time, as well as the communication, file read, and file write bandwidths were each captured across multiple scenarios to evaluate the overall system performance and scalability. Results indicate the implementation scales well—with a 128x increase in the number of parallel processes, the system executed 200x faster. Further, the communication time at its peak accounted for only 12% of the total processing time, indicating the implementation is currently CPU bound and thus should continue to scale well across more and more nodes.
منابع مشابه
A PARCUBE: Sparse Parallelizable CANDECOMP-PARAFAC Tensor Decomposition
How can we efficiently decompose a tensor into sparse factors, when the data does not fit in memory? Tensor decompositions have gained a steadily increasing popularity in data mining applications, however the current state-of-art decomposition algorithms operate on main memory and do not scale to truly large datasets. In this work, we propose PARCUBE, a new and highly parallelizable method for ...
متن کاملA Medium-Grained Algorithm for Distributed Sparse Tensor Factorization
Modeling multi-way data can be accomplished using tensors, which are data structures indexed along three or more dimensions. Tensors are increasingly used to analyze extremely large and sparse multi-way datasets in life sciences, engineering, and business. The canonical polyadic decomposition (CPD) is a popular tensor factorization for discovering latent features and is most commonly found via ...
متن کاملTensor Decompositions for Very Large Scale Problems
Modern applications such as neuroscience, text mining, and large-scale social networks generate massive amounts of data with multiple aspects and high dimensionality. Tensors (i.e., multi-way arrays) provide a natural representation for such massive data. Consequently, tensor decompositions and factorizations are emerging as novel and promising tools for exploratory analysis of multidimensional...
متن کاملDinTucker: Scaling up Gaussian process models on multidimensional arrays with billions of elements
Infinite Tucker Decomposition (InfTucker) and random function prior models, as nonparametric Bayesian models on infinite exchangeable arrays, are more powerful models than widely-used multilinear factorization methods including Tucker and PARAFAC decomposition, (partly) due to their capability of modeling nonlinear relationships between array elements. Despite their great predictive performance...
متن کاملConsensus-based In-Network Computation of the PARAFAC Decomposition
Higher-order tensor analysis is a multi-disciplinary tool widely used in numerous application areas involving data analysis such as psychometrics, chemometrics, and signal processing, just to mention a few. The parallel factor (PARAFAC) decomposition, also known by the acronym CP (standing for “CANDECOMP/PARAFAC” or yet “canonical polyadic”) is the most popular tensor decomposition. Its widespr...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016